STDPboost: A Self-Training Method Based on Density Peaks and Improved Adaboost for Semi-Supervised Classification

نویسندگان

چکیده

The self-training methods have been praised by extensive research in semi-supervised classification. Mislabeling is the main challenge methods. Multiple variations of are recently proposed against mislabeling from following one two aspects: a) using heuristic rules to find high-confidence unlabeled samples that can easily be predicted correctly each iteration; b) enhancing prediction performance employing ensemble classifiers composed multiple weak classifiers. Yet, they still suffer issues: a); most strategies for finding heavily rely on parameters; almost all employed originally designed supervised and may not suitable classification due limited number unrepresented distribution initial labeled data; c) few overcome above aspects at same time. To advance state art, a new method based density peaks clustering improved Adaboost presented named as STDPboost. In iterative self-taught process, clustering-based strategy classifier AdaboostSEMI more predict samples, which overcomes mentioned shortcomings existing Intensive experiments benchmark data sets proven STDPboost outperforms 7 state-of-the-art average accuracy KNN CART with percentages 10% 50% further alleviating mislabeling.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Semi-supervised Affinity Propagation Based on Density Peaks

Original scientific paper In view of the unsatisfying clustering effect of affinity propagation (AP) clustering algorithm when dealing with data sets of complex structures, a semi-supervised affinity propagation clustering algorithm based on density peaks (SAP-DP) was proposed in this paper. The algorithm uses a new algorithm of density peaks (DP) which has the advantage of the manifold cluster...

متن کامل

A Self-training Method for Semi-supervised Gans

Since the creation of Generative Adversarial Networks (GANs), much work has been done to improve their training stability, their generated image quality, their range of application but nearly none of them explored their self-training potential. Self-training has been used before the advent of deep learning in order to allow training on limited labelled training data and has shown impressive res...

متن کامل

A Self-Training Method for Semi-Supervised GANs

Since the creation of Generative Adversarial Networks (GANs), much work has been done to improve their training stability, their generated image quality, their range of application but nearly none of them explored their self-training potential. Self-training has been used before the advent of deep learning in order to allow training on limited labelled training data and has shown impressive res...

متن کامل

A Self-training Method for Semi-supervised Gans

Since the creation of Generative Adversarial Networks (GANs), much work has been done to improve their training stability, their generated image quality, their range of application but nearly none of them explored their self-training potential. Self-training has been used before the advent of deep learning in order to allow training on limited labelled training data and has shown impressive res...

متن کامل

Semi-supervised Self-training for Sentence Subjectivity Classification

Recent natural language processing (NLP) research shows that identifying and extracting subjective information from texts can benefit many NLP applications. In this paper, we address a semi-supervised learning approach, self-training, for sentence subjectivity classification. In self-training, the confidence degree that depends on the ranking of class membership probabilities is commonly used a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2023

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2023.3294982